18 research outputs found

    Explaining Trained Neural Networks with Semantic Web Technologies: First Steps

    Get PDF
    The ever increasing prevalence of publicly available structured data on the World Wide Web enables new applications in a variety of domains. In this paper, we provide a conceptual approach that leverages such data in order to explain the input-output behavior of trained artificial neural networks. We apply existing Semantic Web technologies in order to provide an experimental proof of concept

    Explaining Deep Learning Hidden Neuron Activations using Concept Induction

    Full text link
    One of the current key challenges in Explainable AI is in correctly interpreting activations of hidden neurons. It seems evident that accurate interpretations thereof would provide insights into the question what a deep learning system has internally \emph{detected} as relevant on the input, thus lifting some of the black box character of deep learning systems. The state of the art on this front indicates that hidden node activations appear to be interpretable in a way that makes sense to humans, at least in some cases. Yet, systematic automated methods that would be able to first hypothesize an interpretation of hidden neuron activations, and then verify it, are mostly missing. In this paper, we provide such a method and demonstrate that it provides meaningful interpretations. It is based on using large-scale background knowledge -- a class hierarchy of approx. 2 million classes curated from the Wikipedia Concept Hierarchy -- together with a symbolic reasoning approach called \emph{concept induction} based on description logics that was originally developed for applications in the Semantic Web field. Our results show that we can automatically attach meaningful labels from the background knowledge to individual neurons in the dense layer of a Convolutional Neural Network through a hypothesis and verification process.Comment: Submitted to IJCAI-2

    Neuro-Symbolic Deductive Reasoning for Cross-Knowledge Graph Entailment

    Get PDF
    A significant and recent development in neural-symbolic learning are deep neural networks that can reason over symbolic knowledge graphs (KGs). A particular task of interest is KG entailment, which is to infer the set of all facts that are a logical consequence of current and potential facts of a KG. Initial neural-symbolic systems that can deduce the entailment of a KG have been presented, but they are limited: current systems learn fact relations and entailment patterns specific to a particular KG and hence do not truly generalize, and must be retrained for each KG they are tasked with entailing. We propose a neural-symbolic system to address this limitation in this paper. It is designed as a differentiable end-to-end deep memory network that learns over abstract, generic symbols to discover entailment patterns common to any reasoning task. A key component of the system is a simple but highly effective normalization process for continuous representation learning of KG entities within memory networks. Our results show how the model, trained over a set of KGs, can effectively entail facts from KGs excluded from the training, even when the vocabulary or the domain of test KGs is completely different from the training KGs

    Improved Javanese script recognition using custom model of convolution neural network

    Get PDF
    Handwriting recognition in Javanese script is not widely developed with deep learning (DL). Previous DL and machine learning (ML) research is generally limited to basic characters (Carakan) only. This study proposes a deep learning model using a custom-built convolutional neural network to improve recognition accuracy performance and reduce computational costs. The main features of handwritten objects are textures, edges, lines, and shapes, so convolution layers are not designed in large numbers. This research maximizes optimization of other layers such as pooling, activation function, fully connected layer, optimizer, and parameter settings such as dropout and learning rate. There are eleven main layers used in the proposed custom convolutional neural network (CNN) model, namely four convolution layers+activation function, four pooling layers, two fully connected layers, and a softmax classifier. Based on the test results on the Javanese script handwritten image dataset with 120 classes consisting of 20 basic character classes and 100 compound character classes, the resulting accuracy is 97.29%
    corecore